Skip to content

Conversation

markurtz
Copy link
Collaborator

@markurtz markurtz commented Aug 28, 2024

Summary

Reconfigures tokenizer support for the GuideLLM backend and request generators, enhancing the flexibility and reconfigurability of benchmark requests. It also updates CLI commands to include mandatory arguments for better user guidance and integrates a more comprehensive set of configuration options for data handling and request rate types.

Details

  • Tokenizer Support: Added methods to instantiate and utilize tokenizers in backend classes and request generators, ensuring compatibility with various model configurations.

  • CLI Enhancements:

    • Updated CLI commands to require -data and -data-type arguments, improving clarity for users and preventing misconfigurations.
    • Refined help messages for all CLI options to provide more detailed guidance.
  • Configuration Options:

    • Introduced new options for specifying the -tokenizer and additional request rates in -rate.
    • Added functionality for testing backend connections using tokenizers.
    • Improved error handling when required options or compatible models are not available.
  • Documentation: Updated README.md and added detailed instructions for using the new configuration options.

  • Tests:

    • Expanded unit tests to cover new methods and configurations.
    • Ensured backward compatibility by validating default behaviors with updated test cases.

    Fixes

@markurtz markurtz self-assigned this Aug 28, 2024
@markurtz markurtz merged commit 2f56e0c into main Aug 28, 2024
@markurtz markurtz deleted the readme-adjustments branch August 28, 2024 21:57
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Wrong tokenizer used by default How to specify input/output token lengths for requests?

1 participant